Cerebras-GPT 590M is a language model based on the Transformer architecture and belongs to the Cerebras-GPT model family. It aims to study the scaling laws of large language models and demonstrate the simplicity and scalability of training large language models on the Cerebras software and hardware stack.
Large Language Model
Transformers English